12 research outputs found

    The use of self-organising maps to investigate heat demand profiles

    Get PDF
    District heating companies are responsible for delivering the heat produced in central heat plants to the consumers through a pipeline system. At the same time they are expected to keep the total heat production cost as low as possible. Therefore, there is a growing need to optimise heat production through better prediction of customers needs. The paper illustrates the way neural networks, namely self-organised maps can be used to investigate long-term demand profiles of consumers. Real-life historical sales data is used to establish a number of typical demand profiles

    A Survey on Semi-Supervised Learning for Delayed Partially Labelled Data Streams

    Full text link
    Unlabelled data appear in many domains and are particularly relevant to streaming applications, where even though data is abundant, labelled data is rare. To address the learning problems associated with such data, one can ignore the unlabelled data and focus only on the labelled data (supervised learning); use the labelled data and attempt to leverage the unlabelled data (semi-supervised learning); or assume some labels will be available on request (active learning). The first approach is the simplest, yet the amount of labelled data available will limit the predictive performance. The second relies on finding and exploiting the underlying characteristics of the data distribution. The third depends on an external agent to provide the required labels in a timely fashion. This survey pays special attention to methods that leverage unlabelled data in a semi-supervised setting. We also discuss the delayed labelling issue, which impacts both fully supervised and semi-supervised methods. We propose a unified problem setting, discuss the learning guarantees and existing methods, explain the differences between related problem settings. Finally, we review the current benchmarking practices and propose adaptations to enhance them

    Exploring the 3D Surfaces with Modified Method of Steepest Descent

    No full text
    Aim: To prove expediency of the steepest descent method to divide a given cloud of (Y, X1, X2) points into the spatial clusters with purpose to estimate a simple regression model Y = f(Z|X1,X2) at each cluster. Material and Method: The exemplary data sets {Y, X1, X2} were drawn randomly from assumed 3D surface: Y = f(X1,X2), and then a random noise was added to variable Y. A polynomial model Y = f(X1,X2) and a set of models Y = f(Z|X1,X2) were estimated separately, both under Akaike information criterion (AIC), and then compared with respect to their determination coefficients R-square, and the residuals’ distributions. Results: In the artificial data set studied, the both compared methods after several iterations can provide regression models of the quite similar quality. Conclusions: Because the proposed novel method seems to be more robust to outliers, and easier to graphical presentations and to intuitive understanding than the conventional way of building a regression model, the proposed novel method can be recommended to use by non-statisticians, especially in situation when, besides usual moderate noise, the sporadic but influential measurement errors can occur

    Performance measures for evolving predictions under delayed labelling classification

    Get PDF
    For many streaming classification tasks, the ground truth labels become available with a non-negligible latency. Given this delayed labelling setting, after the instance data arrives and before its true label is known, the online classifier model may change. Hence, the initial prediction can be replaced with additional periodic predictions gradually produced before the true label becomes available. The quality of these predictions may largely vary. Thus, the question arises of how to summarise the performance of these models when multiple predictions for a single instance are made due to delayed labels.In this study, we aim to provide intuitive performance measures summarising the performance of multiple predictions made for individual instances before their true labels arrive. Particular attention is paid to the fact that under the delayed label setting, the emphasis placed on the quality of initial predictions can vary depending on problem needs. The intermediate performance measures we propose complement existing initial and test-then-train performance evaluation when verification latency is observed. Results provided for both real and synthetic datasets show that the new measures can be used to easily rank methods in terms of their ability to produce and refine predictions before the true labels arrive

    Advances in network systems: architectures, security, and applications

    No full text
    This book provides the reader with a comprehensive selection of cutting–edge algorithms, technologies, and applications. The volume offers new insights into a range of fundamentally important topics in network architectures, network security, and network applications. It serves as a reference for researchers and practitioners by featuring research contributions exemplifying research done in the field of network systems. In addition, the book highlights several key topics in both theoretical and practical aspects of networking. These include wireless sensor networks, performance of TCP connections in mobile networks, photonic data transport networks, security policies, credentials management, data encryption for network transmission, risk management, live TV services, and multicore energy harvesting in distributed systems.

    S2CE: A Hybrid Cloud and Edge Orchestrator for Mining Exascale Distributed Streams

    No full text
    The explosive increase in volume, velocity, variety, and veracity of data generated by distributed and heterogeneous nodes such as IoT and other devices, continuously challenge the state of art in big data processing platforms and mining techniques. Consequently, it reveals an urgent need to address the ever-growing gap between this expected exascale data generation and the extraction of insights from these data. To address this need, this paper proposes Stream to Cloud & Edge (S2CE), a first of its kind, optimized, multi-cloud and edge orchestrator, easily configurable, scalable, and extensible. S2CE will enable machine and deep learning over voluminous and heterogeneous data streams running on hybrid cloud and edge settings, while offering the necessary functionalities for practical and scalable processing: data fusion and preprocessing, sampling and synthetic stream generation, cloud and edge smart resource management, and distributed processing

    Modeling pulsed laser micromachining of micro geometries using machine-learning techniques

    No full text
    A wide range of opportunities are emerging in the micro-system technology sector for laser micro-machining systems, because they are capable of processing various types of materials with micro-scale precision. However, few process datasets and machine-learning techniques are optimized for this industrial task. This study describes the process parameters of micro-laser milling and their influence on the final features of the microshapes that are produced. It also identifies the most accurate machine-learning technique for the modelization of this multivariable process. It examines the capabilities of laser micro-machining by performing experiments on hardened steel with a pulsed Nd:YAG laser. Arrays of micro-channels were manufactured using various scanning speeds, pulse intensities and pulse frequencies. The results are presented in terms of the main industrial requirements for any manufactured good: dimensional accuracy (in our case, depth and width of the channels), surface roughness and material removal rate (which is a measure of the productivity of the process). Different machine-learning techniques were then tested on the datasets to build highly accurate models for each output variable. The selected techniques were: k-Nearest Neighbours, neural networks, decision trees and linear regression models. Our analysis of the correlation coefficients and the mean absolute error of all the generated models show that neural networks are better at modelling channel depth and that decision trees are better at modelling material removal rate; both techniques were similar for width and surface roughness. In general, these two techniques show better accuracy than the other two models. The work concludes that decision trees should be used, if information on input parameter relations is sought, while neural networks are suitable when the dimensional accuracy of the workpiece is the main industrial requirement. Extensive datasets are necessary for this industrial task, to provide reliable AI models due to the high rates of noise, especially for some outputs such as roughnes
    corecore